Maximal Margin Classification for Metric Spaces

نویسندگان

  • Matthias Hein
  • Olivier Bousquet
چکیده

In this article we construct a maximal margin classification algorithm for arbitrary metric spaces. At first we show that the Support Vector Machine (SVM) is a maximal margin algorithm for the class of metric spaces where the negative squared distance is conditionally positive definite (CPD). This means that the metric space can be isometrically embedded into a Hilbert space, where one performs linear maximal margin separation. We will show that the solution only depends on the metric, but not on the kernel. Following the framework we develop for the SVM, we construct an algorithm for maximal margin classification in arbitrary metric spaces. The main difference compared with SVM is that we no longer embed isometrically into a Hilbert space, but a Banach space. We further give an estimate of the capacity of the function class involved in this algorithm via Rademacher averages. We recover an algorithm of Graepel et al. [6].

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Multi-granularity distance metric learning via neighborhood granule margin maximization

Learning a distance metric from training samples is often a crucial step in machine learning and pattern recognition. Locality, compactness and consistency are considered as the key principles in distance metric learning. However, the existing metric learning methods just consider one or two of them. In this paper, we develop a multi-granularity distance learning technique. First, a new index, ...

متن کامل

Distance-Based Classification with Lipschitz Functions

The goal of this article is to develop a framework for large margin classification in metric spaces. We want to find a generalization of linear decision functions for metric spaces and define a corresponding notion of margin such that the decision function separates the training points with a large margin. It will turn out that using Lipschitz functions as decision functions, the inverse of the...

متن کامل

W-convergence of the proximal point algorithm in complete CAT(0) metric spaces

‎In this paper‎, ‎we generalize the proximal point algorithm to complete CAT(0) spaces and show‎ ‎that the sequence generated by the proximal point algorithm‎ $w$-converges to a zero of the maximal‎ ‎monotone operator‎. ‎Also‎, ‎we prove that if $f‎: ‎Xrightarrow‎ ‎]-infty‎, +‎infty]$ is a proper‎, ‎convex and lower semicontinuous‎ ‎function on the complete CAT(0) space $X$‎, ‎then the proximal...

متن کامل

A Proximal Point Algorithm for Finding a Common Zero of a Finite Family of Maximal Monotone Operators

In this paper, we consider a proximal point algorithm for finding a common zero of a finite family of maximal monotone operators in real Hilbert spaces. Also, we give a necessary and sufficient condition for the common zero set of finite operators to be nonempty, and by showing that in this case, this iterative sequence converges strongly to the metric projection of some point onto the set of c...

متن کامل

Geometrical aspects of statistical learning theory

Geometry plays an important role in modern statistical learning theory, and many different aspects of geometry can be found in this fast developing field. This thesis addresses some of these aspects. A large part of this work will be concerned with so called manifold methods, which have recently attracted a lot of interest. The key point is that for a lot of real-world data sets it is natural t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • J. Comput. Syst. Sci.

دوره 71  شماره 

صفحات  -

تاریخ انتشار 2003